Ergodic Results in Subgradient Optimization

نویسنده

  • Michael Patriksson
چکیده

Subgradient methods are popular tools for nonsmooth, convex minimization , especially in the context of Lagrangean relaxation; their simplicity has been a main contribution to their success. As a consequence of the nonsmoothness, it is not straightforward to monitor the progress of a subgradient method in terms of the approximate fulllment of optimality conditions, since the subgradients used in the method will, in general, not accumulate to subgradients that verify optimality of a solution obtained in the limit. Further, certain supplementary information, such as convergent estimates of Lagrange multipliers, is not directly available in subgradient schemes. As a means for overcoming these weaknesses of subgradient optimization methods , we introduce the computation of an ergodic (averaged) sequence of subgradients. Speciically, we consider a nonsmooth, convex program solved by a conditional subgra-dient optimization scheme (of which the traditional subgradient optimization method is a special case) with divergent series step lengths, which generates a sequence of iterates that converges to an optimal solution. We show that the elements of the ergodic sequence of subgradients in the limit fulll the optimality conditions at this optimal solution. Further, we use the convergence properties of the ergodic sequence of subgra-dients to establish convergence of an ergodic sequence of Lagrange multipliers. Finally, some potential applications of these ergodic results are brieey discussed.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Ergodic, primal convergence in dual subgradient schemes for convex programming

Lagrangean dualization and subgradient optimization techniques are frequently used within the field of computational optimization for finding approximate solutions to large, structured optimization problems. The dual subgradient scheme does not automatically produce primal feasible solutions; there is an abundance of techniques for computing such solutions (via penalty functions, tangential app...

متن کامل

Ergodic Convergence in Subgradient Optimization

When nonsmooth, convex minimizationproblems are solved by subgradientoptimizationmethods, the subgradients used will in general not accumulate to subgradients which verify the optimal-ity of a solution obtained in the limit. It is therefore not a straightforward task to monitor the progress of a subgradient method in terms of the approximate fulllment of optimality conditions. Further, certain ...

متن کامل

Ergodic, primal convergence in dual subgradient schemes for convex programming, II: the case of inconsistent primal problems

Consider the utilization of a Lagrangian dual method which is convergent for consistent convex optimization problems. When it is used to solve an infeasible optimization problem, its inconsistency will then manifest itself through the divergence of the sequence of dual iterates. Will then the sequence of primal subproblem solutions still yield relevant information regarding the primal program? ...

متن کامل

Technical Report Subgradient Optimization for Convex Multiparametric Programming

In this paper we develop a subgradient optimization methodology for convex multiparametric nonlinear programs. We define a parametric subgradient, extend some classical optimization results to the multiparametric case, and design a subgradient algorithm that is shown to converge under traditional conditions. We use this algorithm to solve two illustrative example problems and demonstrate its ac...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 1995